DyNet: The Dynamic Neural Network Toolkit
نویسندگان
چکیده
We describe DyNet, a toolkit for implementing neural network models based on dynamic declaration of network structure. In the static declaration strategy that is used in toolkits like Theano, CNTK, and TensorFlow, the user first defines a computation graph (a symbolic representation of the computation), and then examples are fed into an engine that executes this computation and computes its derivatives. In DyNet’s dynamic declaration strategy, computation graph construction is mostly transparent, being implicitly constructed by executing procedural code that computes the network outputs, and the user is free to use different network structures for each input. Dynamic declaration thus facilitates the implementation of more complicated network architectures, and DyNet is specifically designed to allow users to implement their models in a way that is idiomatic in their preferred programming language (C++ or Python). One challenge with dynamic declaration is that because the symbolic computation graph is defined anew for every training example, its construction must have low overhead. To achieve this, DyNet has an optimized C++ backend and lightweight graph representation. Experiments show that DyNet’s speeds are faster than or comparable with static declaration toolkits, and significantly faster than Chainer, another dynamic declaration toolkit. DyNet is released opensource under the Apache 2.0 license, and available at http://github.com/clab/dynet. Carnegie Mellon University, Pittsburgh, PA, USA Nara Institute of Science and Technology, Ikoma, Japan DeepMind, London, UK Bar Ilan University, Ramat Gan, Israel Allen Institute for Artificial Intelligence, Seattle, WA, USA University of Notre Dame, Notre Dame, IN, USA IBM T.J. Watson Research Center, Yorktown Heights, NY, USA University of Melbourne, Melbourne, Australia Johns Hopkins University, Baltimore, MD, USA Google, New York, NY, USA Google, Mountain View, CA, USA University of Washington, Seattle, USA Microsoft Research, Redmond, WA, USA University of Edinburgh, Edinburgh, UK 1 ar X iv :1 70 1. 03 98 0v 1 [ st at .M L ] 1 5 Ja n 20 17
منابع مشابه
On-the-fly Operation Batching in Dynamic Computation Graphs
Dynamic neural network toolkits such as PyTorch, DyNet, and Chainer offer more flexibility for implementing models that cope with data of varying dimensions and structure, relative to toolkits that operate on statically declared computations (e.g., TensorFlow, CNTK, and Theano). However, existing toolkits—both static and dynamic—require that the developer organize the computations into the batc...
متن کاملDyNet: visualization and analysis of dynamic molecular interaction networks
UNLABELLED : The ability to experimentally determine molecular interactions on an almost proteome-wide scale under different conditions is enabling researchers to move from static to dynamic network analysis, uncovering new insights into how interaction networks are physically rewired in response to different stimuli and in disease. Dynamic interaction data presents a special challenge in netwo...
متن کاملCavs: A Vertex-centric Programming Interface for Dynamic Neural Networks
Recent deep learning (DL) models have moved beyond static network architectures to dynamic ones, handling data where the network structure changes every example, such as sequences of variable lengths, trees, and graphs. Existing dataflow-based programming models for DL—both static and dynamic declaration—either cannot readily express these dynamic models, or are inefficient due to repeated data...
متن کاملComparative Study of Static and Dynamic Artificial Neural Network Models in Forecasting of Tehran Stock Exchange
During the recent decades, neural network models have been focused upon by researchers due to their more real performance and on this basis, different types of these models have been used in forecasting. Now, there is a question that which kind of these models has more explanatory power in forecasting the future processes of the stock. In line with this, the present paper made a comparison betw...
متن کاملA Convolutional Neural Network based on Adaptive Pooling for Classification of Noisy Images
Convolutional neural network is one of the effective methods for classifying images that performs learning using convolutional, pooling and fully-connected layers. All kinds of noise disrupt the operation of this network. Noise images reduce classification accuracy and increase convolutional neural network training time. Noise is an unwanted signal that destroys the original signal. Noise chang...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- CoRR
دوره abs/1701.03980 شماره
صفحات -
تاریخ انتشار 2017